301 research outputs found
Differentially Private Publication of Sparse Data
The problem of privately releasing data is to provide a version of a dataset
without revealing sensitive information about the individuals who contribute to
the data. The model of differential privacy allows such private release while
providing strong guarantees on the output. A basic mechanism achieves
differential privacy by adding noise to the frequency counts in the contingency
tables (or, a subset of the count data cube) derived from the dataset. However,
when the dataset is sparse in its underlying space, as is the case for most
multi-attribute relations, then the effect of adding noise is to vastly
increase the size of the published data: it implicitly creates a huge number of
dummy data points to mask the true data, making it almost impossible to work
with.
We present techniques to overcome this roadblock and allow efficient private
release of sparse data, while maintaining the guarantees of differential
privacy. Our approach is to release a compact summary of the noisy data.
Generating the noisy data and then summarizing it would still be very costly,
so we show how to shortcut this step, and instead directly generate the summary
from the input data, without materializing the vast intermediate noisy data. We
instantiate this outline for a variety of sampling and filtering methods, and
show how to use the resulting summary for approximate, private, query
answering. Our experimental study shows that this is an effective, practical
solution, with comparable and occasionally improved utility over the costly
materialization approach
One-loop expressions for in Higgs extensions of the Standard Model
A systematic study of one-loop contributions to the decay channels
with , performed in
Higgs extended versions of the Standard Model, is presented in the 't
Hooft-Veltman gauge. Analytic formulas for one-loop form factors are expressed
in terms of the logarithm and di-logarithmic functions. As a result, these form
factors can be reduced to those relating to the loop-induced decay processes
, confirming not only previous results
using different approaches but also close relations between the three kinds of
the loop-induced Higgs decay rates. For phenomenological study, we focus on the
two observables, namely the enhancement factors defined as ratios of the decay
rates calculated between the Higgs extended versions and the standard model,
and the forward-backward asymmetries of fermions, which can be used to search
for Higgs extensions of the SM. We show that direct effects of mixing between
neutral Higgs bosons and indirect contributions of charged Higg boson exchanges
can be probed at future colliders.Comment: 39 pages, 9 Figures, 11 Tables of dat
Assessment of seasonal winter temperature forecast errors in the regcm model over northern Vietnam
This study verified the seasonal six-month forecasts for winter temperatures for northern Vietnam in 1998–2018 using a regional climate model (RegCM4) with the boundary conditions of the climate forecast system Version 2 (CFSv2) from the National Centers for Environmental Prediction (NCEP). First, different physical schemes (land-surface process, cumulus, and radiation parameterizations) in RegCM4 were applied to generate 12 single forecasts. Second, the simple ensemble forecasts were generated through the combinations of those different physical formulations. Three subclimate regions (R1, R2, R3) of northern Vietnam were separately tested with surface observations and a reanalysis dataset (Japanese 55-year reanalysis (JRA55)). The highest sensitivity to the mean monthly temperature forecasts was shown by the land-surface parameterizations (the biosphere−atmosphere transfer scheme (BATS) and community land model version 4.5 (CLM)). The BATS forecast groups tended to provide forecasts with lower temperatures than the actual observations, while the CLM forecast groups tended to overestimate the temperatures. The forecast errors from single forecasts could be clearly reduced with ensemble mean forecasts, but ensemble spreads were less than those root-mean-square errors (RMSEs). This indicated that the ensemble forecast was underdispersed and that the direct forecast from RegCM4 needed more postprocessing
Seasonal variation of phytoplankton in My Thanh River, Mekong delta, Vietnam
A study on the seasonal variation of phytoplankton composition was conducted at the upper, middle, and lower parts of the My Thanh River, which supplies an important source of water for aquaculture. Qualitative and quantitative samples of phytoplankton were collected monthly at both high and low tide. The results showed that a total of 171 phytoplankton (algae) species were recorded, belonging to 59 genera and 5 phyla. Diatoms were the most abundant group with the highest species number, followed by green algae. The other phyla possessed a lower number of species. The species composition was more diverse in the rainy season and at high tide at most of the sampling sites. The mean density of algae varied from 30,900-43,521 ind.L^-1^. The density of diatoms was higher in the middle and lower parts. At the same time, euglenoids displayed the highest density in the upper part, showing a difference in the dominant algae group under the influence of salinity. Salinity was found to be significantly positively correlated (p<0.01) with diatoms, whereas it was negatively correlated (p<0.05) with blue-green algae and euglenoids. The algae composition was quite diverse, with the H' index ranging from 2.0-3.3, showing the water quality was slightly to moderately polluted
Discovering Valuable Items from Massive Data
Suppose there is a large collection of items, each with an associated cost
and an inherent utility that is revealed only once we commit to selecting it.
Given a budget on the cumulative cost of the selected items, how can we pick a
subset of maximal value? This task generalizes several important problems such
as multi-arm bandits, active search and the knapsack problem. We present an
algorithm, GP-Select, which utilizes prior knowledge about similarity be- tween
items, expressed as a kernel function. GP-Select uses Gaussian process
prediction to balance exploration (estimating the unknown value of items) and
exploitation (selecting items of high value). We extend GP-Select to be able to
discover sets that simultaneously have high utility and are diverse. Our
preference for diversity can be specified as an arbitrary monotone submodular
function that quantifies the diminishing returns obtained when selecting
similar items. Furthermore, we exploit the structure of the model updates to
achieve an order of magnitude (up to 40X) speedup in our experiments without
resorting to approximations. We provide strong guarantees on the performance of
GP-Select and apply it to three real-world case studies of industrial
relevance: (1) Refreshing a repository of prices in a Global Distribution
System for the travel industry, (2) Identifying diverse, binding-affine
peptides in a vaccine de- sign task and (3) Maximizing clicks in a web-scale
recommender system by recommending items to users
Resource Competition on Integral Polymatroids
We study competitive resource allocation problems in which players distribute
their demands integrally on a set of resources subject to player-specific
submodular capacity constraints. Each player has to pay for each unit of demand
a cost that is a nondecreasing and convex function of the total allocation of
that resource. This general model of resource allocation generalizes both
singleton congestion games with integer-splittable demands and matroid
congestion games with player-specific costs. As our main result, we show that
in such general resource allocation problems a pure Nash equilibrium is
guaranteed to exist by giving a pseudo-polynomial algorithm computing a pure
Nash equilibrium.Comment: 17 page
Conquering hypertension in Vietnam-solutions at grassroots level: study protocol of a cluster randomized controlled trial
BACKGROUND: Vietnam has been experiencing an epidemiologic transition to that of a lower-middle income country with an increasing prevalence of non-communicable diseases. The key risk factors for cardiovascular disease (CVD) are either on the rise or at alarming levels in Vietnam, particularly hypertension (HTN). Inasmuch, the burden of CVD will continue to increase in the Vietnamese population unless effective prevention and control measures are put in place. The objectives of the proposed project are to evaluate the implementation and effectiveness of two multi-faceted community and clinic-based strategies on the control of elevated blood pressure (BP) among adults in Vietnam via a cluster randomized trial design.
METHODS: Sixteen communities will be randomized to either an intervention (8 communities) or a comparison group (8 communities). Eligible and consenting adult study participants with HTN (n = 680) will be assigned to intervention/comparison status based on the community in which they reside. Both comparison and intervention groups will receive a multi-level intervention modeled after the Vietnam National Hypertension Program including education and practice change modules for health care providers, accessible reading materials for patients, and a multi-media community awareness program. In addition, the intervention group only will receive three carefully selected enhancements integrated into routine clinical care: (1) expanded community health worker services, (2) home BP self-monitoring, and (3) a storytelling intervention, which consists of interactive, literacy-appropriate, and culturally sensitive multi-media storytelling modules for motivating behavior change through the power of patients speaking in their own voices. The storytelling intervention will be delivered by DVDs with serial installments at baseline and at 3, 6, and 9 months after trial enrollment. Changes in BP will be assessed in both groups at several follow-up time points. Implementation outcomes will be assessed as well.
DISCUSSION: Results from this full-scale trial will provide health policymakers with practical evidence on how to combat a key risk factor for CVD using a feasible, sustainable, and cost-effective intervention that could be used as a national program for controlling HTN in Vietnam.
TRIAL REGISTRATION: ClinicalTrials.gov NCT03590691 . Registered on July 17, 2018. Protocol version: 6. Date: August 15, 2019
Solar models and solar neutrino oscillations
We provide a summary of the current knowledge, theoretical and experimental,
of solar neutrino fluxes and of the masses and mixing angles that characterize
solar neutrino oscillations. We also summarize the principal reasons for doing
new solar neutrino experiments and what we think may be learned from the future
measurements.Comment: Submitted to the Neutrino Focus Issue of New Journal of Physics at
http://www.njp.or
Mapping for engagement: setting up a community based participatory research project to reach underserved communities at risk for Hepatitis C in Ho Chi Minh City, Vietnam
Background: Approximately 1. 07 million people in Vietnam are infected with hepatitis C virus (HCV). To address this epidemic, the South East Asian Research Collaborative in Hepatitis (SEARCH) launched a 600-patient cohort study and two clinical trials, both investigating shortened treatment strategies for chronic HCV infection with direct-acting antiviral drugs. We conducted ethnographic research with a subset of trial participants and found that the majority were aware of HCV infection and its implications and were motivated to seek treatment. However, people who inject drugs (PWID), and other groups at risk for HCV were under-represented, although injecting drug use is associated with high rates of HCV. Material and Methods: We designed a community-based participatory research (CBPR) study to engage in dialogues surrounding HCV and other community-prioritized health issues with underserved groups at risk for HCV in Ho Chi Minh City. The project consists of three phases: situation analysis, CBPR implementation, and dissemination. In this paper, we describe the results of the first phase (i.e., the situation analysis) in which we conducted desk research and organized stakeholder mapping meetings with representatives from local non-government and community-based organizations where we used participatory research methods to identify and analyze key stakeholders working with underserved populations. Results: Twenty six institutions or groups working with the key underserved populations were identified. Insights about the challenges and dynamics of underserved communities were also gathered. Two working groups made up of representatives from the NGO and CBO level were formed. Discussion: Using the information provided by local key stakeholders to shape the project has helped us to build solid relationships, give the groups a sense of ownership from the early stages, and made the project more context specific. These steps are not only important preliminary steps for participatory studies but also for other research that takes place within the communities
- …